Learning on Graph with Laplacian Regularization

نویسندگان

  • Rie Kubota Ando
  • Tong Zhang
چکیده

We consider a general form of transductive learning on graphs with Laplacian regularization, and derive margin-based generalization bounds using appropriate geometric properties of the graph. We use this analysis to obtain a better understanding of the role of normalization of the graph Laplacian matrix as well as the effect of dimension reduction. The results suggest a limitation of the standard degree-based normalization. We propose a remedy from our analysis and demonstrate empirically that the remedy leads to improved classification performance.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

1 Discrete Regularization

Many real-world machine learning problems are situated on finite discrete sets, including dimensionality reduction, clustering, and transductive inference. A variety of approaches for learning from finite sets has been proposed from different motivations and for different problems. In most of those approaches, a finite set is modeled as a graph, in which the edges encode pairwise relationships ...

متن کامل

On the Effectiveness of Laplacian Normalization for Graph Semi-supervised Learning

This paper investigates the effect of Laplacian normalization in graph-based semi-supervised learning. To this end, we consider multi-class transductive learning on graphs with Laplacian regularization. Generalization bounds are derived using geometric properties of the graph. Specifically, by introducing a definition of graph cut from learning theory, we obtain generalization bounds that depen...

متن کامل

Graph-Based Posterior Regularization for Semi-Supervised Structured Prediction

We present a flexible formulation of semisupervised learning for structured models, which seamlessly incorporates graphbased and more general supervision by extending the posterior regularization (PR) framework. Our extension allows for any regularizer that is a convex, differentiable function of the appropriate marginals. We show that surprisingly, non-linearity of such regularization does not...

متن کامل

Lecture 6: Manifold Regularization

We first analyze the limits of learning in high dimension. Hence, we stress the difference between high dimensional ambient space and intrinsic geometry associated to the marginal distribution. We observe that, in the semi-supervised setting, unlabeled data could be used to exploit low dimensionality of the intrinsic geometry. In order to formalize these intuitions we briefly introduce the mani...

متن کامل

Regularized Semi-supervised Classification on Manifold

Semi-supervised learning gets estimated marginal distribution X P with a large number of unlabeled examples and then constrains the conditional probability ) | ( x y p with a few labeled examples. In this paper, we focus on a regularization approach for semi-supervised classification. The label information graph is first defined to keep the pairwise label relationship and can be incorporated wi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006